Learning Nonlinear Dynamical Systems Using an EM Algorithm

نویسندگان

  • Zoubin Ghahramani
  • Sam T. Roweis
چکیده

The Expectation{Maximization (EM) algorithm is an iterative procedure for maximum likelihood parameter estimation from data sets with missing or hidden variables [2]. It has been applied to system identi cation in linear stochastic state-space models, where the state variables are hidden from the observer and both the state and the parameters of the model have to be estimated simultaneously [9]. We present a generalization of the EM algorithm for parameter estimation in nonlinear dynamical systems. The \expectation" step makes use of Extended Kalman Smoothing to estimate the state, while the \maximization" step re-estimates the parameters using these uncertain state estimates. In general, the nonlinear maximization step is di cult because it requires integrating out the uncertainty in the states. However, if Gaussian radial basis function (RBF) approximators are used to model the nonlinearities, the integrals become tractable and the maximization step can be solved via systems of linear equations. 1 Stochastic Nonlinear Dynamical Systems We examine inference and learning in discrete-time dynamical systems with hidden state xt, inputs ut, and outputs yt. 1 The state evolves according to stationary nonlinear dynamics driven by the inputs and by additive noise xt+1 = f(xt; ut) + w (1) All lowercase characters (except indices) denote vectors. Matrices are represented by uppercase characters. where w is zero-mean Gaussian noise with covariance Q. 2 The outputs are nonlinearly related to the states and inputs by yt = g(xt; ut) + v (2) where v is zero-mean Gaussian noise with covariance R. The vector-valued nonlinearities f and g are assumed to be di erentiable, but otherwise arbitrary. Models of this kind have been examined for decades in various communities. Most notably, nonlinear state-space models form one of the cornerstones of modern systems and control engineering. In this paper, we examine these models within the framework of probabilistic graphical models and derive a novel learning algorithm for them based on EM. With one exception, this is to the best of our knowledge the rst paper addressing learning of stochastic nonlinear dynamical systems of the kind we have described within the framework of the EM algorithm. The classical approach to system identi cation treats the parameters as hidden variables, and applies the Extended Kalman Filtering algorithm (described in section 2) to the nonlinear system with the state vector augmented by the parameters [5]. This approach is inherently on-line, which may be important in certain applications. Furthermore, it provides an estimate of the covariance of the parameters at each time step. In contrast, the EM algorithm we present is a batch algorithm and does not attempt to estimate the covariance of the parameters. There are three important advantages the EM algorithm has over the classical approach. First, the EM algorithm provides a straightforward and principled method for handing missing inputs or outputs. Second, EM generalizes readily to more complex models with combinations of discrete and real-valued hidden variables. For example, one can formulate EM for a mixture of nonlinear dynamical systems. Third, whereas it is often very di cult to prove or analyze stability within the classical on-line approach, the EM algorithm is always attempting to maximize the likelihood, which acts as a Lyapunov function for stable learning. In the next sections we will describe the basic components of the learning algorithm. For the expectation step of the algorithm, we infer the conditional distribution of the hidden states using Extended Kalman Smoothing (section 2). For the maximization step we rst discuss the general case (section 3) and then describe the particular case where the nonlinearities are represented using Gaussian radial basis function (RBF; [6]) networks (section 4). 2 Extended Kalman Smoothing Given a system described by equations (1) and (2), we need to infer the hidden states from a history of observed inputs and outputs. The quantity at the heart of this inference problem is the conditional density P (xtju1; : : : ; uT ; y1; : : : ; yT ), for 1 t T , which captures the fact that the system is stochastic and therefore our inferences about x will be uncertain. The Gaussian noise assumption is less restrictive for nonlinear systems than for linear systems since the nonlinearity can be used to generate non-Gaussian state noise. The authors have just become aware that Briegel and Tresp (this volume) have applied EM to essentially the same model. Briegel and Tresp's method uses multilayer perceptrons (MLP) to approximate the nonlinearities, and requires sampling from the hidden states to t the MLP. We use Gaussian radial basis functions (RBFs) to model the nonlinearities, which can be t analytically without sampling (see section 4). It is important not to confuse this use of the Extended Kalman algorithm, to simultaneously estimate parameters and hidden states, with our use of EKS, to estimate just the hidden state as part of the E step of EM. For linear dynamical systems with Gaussian state evolution and observation noises, this conditional density is Gaussian and the recursive algorithm for computing its mean and covariance is known as Kalman smoothing [4, 8]. Kalman smoothing is directly analogous to the forward{backward algorithm for computing the conditional hidden state distribution in a hidden Markov model, and is also a special case of the belief propagation algorithm. For nonlinear systems this conditional density is in general non-Gaussian and can in fact be quite complex. Multiple approaches exist for inferring the hidden state distribution of such nonlinear systems, including sampling methods [7] and variational approximations [3]. We focus instead in this paper on a classic approach from engineering, Extended Kalman Smoothing (EKS). Extended Kalman Smoothing simply applies Kalman smoothing to a local linearization of the nonlinear system. At every point ~ x in x-space, the derivatives of the vector-valued functions f and g de ne the matrices, A~ x @f @x x=~ x and C~ x @g @x x=~ x , respectively. The dynamics are linearized about x̂t, the mean of the Kalman lter state estimate at time t: xt+1 = f(x̂t; ut) +Ax̂t (xt x̂t) + w: (3) The output equation (2) can be similarly linearized. If the prior distribution of the hidden state at t = 1 was Gaussian, then, in this linearized system, the conditional distribution of the hidden state at any time t given the history of inputs and outputs will also be Gaussian. Thus, Kalman smoothing can be used on the linearized system to infer this conditional distribution (see gure 1, left panel).

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Determination of Stability Domains for Nonlinear Dynamical Systems Using the Weighted Residuals Method

Finding a suitable estimation of stability domain around stable equilibrium points is an important issue in the study of nonlinear dynamical systems. This paper intends to apply a set of analytical-numerical methods to estimate the region of attraction for autonomous nonlinear systems. In mechanical and structural engineering, autonomous systems could be found in large deformation problems or c...

متن کامل

Adaptive Approximation-Based Control for Uncertain Nonlinear Systems With Unknown Dead-Zone Using Minimal Learning Parameter Algorithm

This paper proposes an adaptive approximation-based controller for uncertain strict-feedback nonlinear systems with unknown dead-zone nonlinearity. Dead-zone constraint is represented as a combination of a linear system with a disturbance-like term. This work invokes neural networks (NNs) as a linear-in-parameter approximator to model uncertain nonlinear functions that appear in virtual and act...

متن کامل

The Study of Nonlinear Dynamical Systems Nuclear Fission Using Hurwitz Criterion

In this paper, the nonlinear dynamic system of equations, a type of nuclear ssion reactor is solved analytically and numerically. Considering that the direct solution of three-dimensional dynamical systems analysis and more in order to determine the stability and instability, in terms of algebraicsystems is dicult. Using certain situations in mathematics called Hurwitz criterion, Necessary and ...

متن کامل

Solving Second Kind Volterra-Fredholm Integral Equations by Using Triangular Functions (TF) and Dynamical Systems

The method of triangular functions (TF) could be a generalization form of the functions of block-pulse (Bp)‎. ‎The solution of second kind integral equations by using the concept of TF would lead to a nonlinear equations system‎. ‎In this article‎, ‎the obtained nonlinear system has been solved as a dynamical system‎. ‎The solution of the obtained nonlinear system by the dynamical system throug...

متن کامل

UW CSE Technical Report 03-06-02 The Trajectory Mixture Model for Learning Collections of Nonlinear Functions

Learning statistical models of nonlinear dynamical systems has long been an important problem in machine learning. The problem becomes especially hard when the dynamical system is composed of a mixture of nonlinear models, not just a single nonlinear model. To address this general case of nonlinear time-series modeling, we propose a new hierarchical architecture: the Trajectory Mixture Model (T...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1998